Not All Dialogues are Created Equal: Instance Weighting for Neural Conversational Models
نویسندگان
چکیده
Neural conversational models require substantial amounts of dialogue data to estimate their parameters and are therefore usually learned on large corpora such as chat forums, Twitter discussions or movie subtitles. These corpora are, however, often challenging to work with, notably due to their frequent lack of turn segmentation and the presence of multiple references external to the dialogue itself. This paper shows that these challenges can be mitigated by adding a weighting model into the neural architecture. The weighting model, which is itself estimated from dialogue data, associates each training example to a numerical weight that reflects its intrinsic quality for dialogue modelling. At training time, these sample weights are included into the empirical loss to be minimised. Evaluation results on retrieval-based models trained on movie and TV subtitles demonstrate that the inclusion of such a weighting model improves the model performance on unsupervised metrics.
منابع مشابه
Deep Reinforcement Learning for Dialogue Generation
Recent neural models of dialogue generation offer great promise for generating responses for conversational agents, but tend to be shortsighted, predicting utterances one at a time while ignoring their influence on future outcomes. Modeling the future direction of a dialogue is crucial to generating coherent, interesting dialogues, a need which led traditional NLP models of dialogue to draw on ...
متن کاملActed and Spontaneous Conversational Prosody — Same or Different?
Previous work has shown that read and spontaneous monologues differ prosodically both in production and perception. In this paper, we examine whether similar effects can be found between spontaneous and read, or rather acted, dialogues. It is possible that speakers can mimic conversational prosody very well. Alternatively, they might use prosodic resources more than the conversational situation...
متن کاملA Long Short-Term Memory Framework for Predicting Humor in Dialogues
We propose a first-ever attempt to employ a Long Short-Term memory based framework to predict humor in dialogues. We analyze data from a popular TV-sitcom, whose canned laughters give an indication of when the audience would react. We model the setuppunchline relation of conversational humor with a Long Short-Term Memory, with utterance encodings obtained from a Convolutional Neural Network. Ou...
متن کاملRe-enacted and Spontaneous Conversational Prosody — How Different?
Previous work has shown that read and spontaneous monologues differ prosodically both in production and perception. In this paper, we examine to which extent similar effects can be found between spontaneous and read, or rather re-enacted, dialogues. It is possible that speakers can mimic conversational prosody very well. Another possibility is that in re-enacted dialogues, prosody is actually u...
متن کاملGricean Brainwaves: Brain Responses to Pragmatic Violations in Dialogues
During conversation, speakers and listeners act on certain basic assumptions, which enable them to communicate swiftly and seemingly effortlessly (Grice, 1975). The speaker, for instance, is supposed to say no more, but also no less than is necessary in a given conversational context (Maxim of Quantity). The present study looks at how language users react when this pragmatic assumption is viola...
متن کامل